Numerical Issues Involved in Inverting Hessian Matrices
نویسندگان
چکیده
In the social sciences, researchers typically assume the accuracy of generalized linear models by using an asymptotic normal approximation to the likelihood function or, occasionally, by using the full posterior distribution. Thus, for standard maximum likelihood analyses, only point estimates and the variance at the maximum are normally seen as necessary. For Bayesian posterior analysis, the maximum and variance provide a useful first approximation (but see Chapter 4 for an alternative). Unfortunately, although the negative of the Hessian (the matrix of second derivatives of the posterior with respect to the parameters and named for its inventor in slightly different context, German mathematician Ludwig Hesse) must be positive definite and hence invertible so as to compute the variance matrix, invertible Hessians do not exist for some combinations of datasets and models, so statistical procedures sometimes fail for this reason before completion. Indeed, receiving a computer-generated "Hessian not invertible" message (because of singularity or nonpositive definiteness) rather than a set of statistical results is a frustrating but common occurrence in applied quantitative research. It even occurs with regularity during many Monte Carlo experiments where the investigator is drawing data from a known statistical model, due to machine effects. The Hessian can be noninvertible for both computational reasons and data reasons. Inaccurate implementation of the likelihood function (see Chapters 2 and 3), inaccurate derivative methods (see Chapter 8), or other inappropriate choices in optimization algorithms can yield noninvertible Hessians. Where these inaccuracies cause problems with Hessians, we recommend addressing these inaccuracies directly. If these methods aren't feasible, or don't work, which often happens, we provide an innovative new library for doing generalized inverses . Moreover, when a Hessian is not invertible for data reasons, no computational trick can make it invertible, given the model and data chosen, because the desired inverse does
منابع مشابه
A Hessian-Based Method for Uncertainty Quantification in Global Ocean State Estimation
Derivative-based methods are developed for uncertainty quantification (UQ) in largescale ocean state estimation. The estimation system is based on the adjoint method for solving a least-squares optimization problem, whereby the state-of-the-art MIT general circulation model (MITgcm) is fit to observations. The UQ framework is applied to quantify Drake Passage transport uncertainties in a global...
متن کاملMatrix Probing and its Conditioning
When a matrix A with n columns is known to be well-approximated by a linear combination of basis matrices B1, . . . , Bp, we can apply A to a random vector and solve a linear system to recover this linear combination. The same technique can be used to obtain an approximation to A−1. A basic question is whether this linear system is well-conditioned. This is important for two reasons: a well-con...
متن کاملConvergence of Rump’s Method for Inverting Arbitrarily Ill-Conditioned Matrices
In this paper, the problem of inverting regular matrices with arbitrarily large condition number is treated in double precision defined by IEEE 754 floating point standard. In about 1984, Rump derived a method for inverting arbitrarily ill-conditioned matrices. The method requires the possibility to calculate a dot product in higher precision. Rump’s method is of theoretical interest. Rump made...
متن کاملVariable-storage quasi-Newton operators as inverse forecast/analysis error covariance matrices in variational data assimilation
Two approximations of the Hessian matrix as limited-memory operators are built from the limited-memoryBFGS inverse Hessian approximationprovided by the minimization code, in view of the speci cation of the inverse analysis/forecast error covariance matrix in variational data assimilation. Some numerical experiments and theoretical considerations lead to reject the limited-memory DFP Hessian app...
متن کاملApproximate inversion of the wave-equation Hessian via randomized matrix probing
We present a method for approximately inverting the Hessian of full waveform inversion as a dip-dependent and scaledependent amplitude correction. The terms in the expansion of this correction are determined by least-squares fitting from a handful of applications of the Hessian to random models — a procedure called matrix probing. We show numerical indications that randomness is important for g...
متن کامل